[Salon] Misinformation research provides an academic veneer for political propaganda



Stone of Madness

Misinformation research provides an academic veneer for political propaganda

The Surgeon, or The Extraction of the Stone, Jan Sanders van Hemessen, 1550-1550. Source: Wikipedia. The original painting is in the Museo del Prado in Madrid, Spain.

The painting above, by Dutch painter Jan Sanders van Hemessen in the 16th century, depicts a man undergoing a rather violent procedure to remove a stone from his brain via a procedure called trephination — the intentional creation of a small hole in the skull. Trephination to remove a “stone of madness” from the head was practiced in medieval times as a supposed cure for “madness, idiocy or dementia”, typically by “medical quacks [who] roamed the countryside offering to perform the surgery.”

Madness, trephination, and self-appointed surgeons roaming the countryside looking for patients of course has me thinking about the recent attention to the academic subfield of misinformation studies. Today, I introduce several arguments about misinformation studies that I’ll be developing in depth here at THB in coming weeks and months as I work on my new book.

They are:

In January, the World Economic Forum warned that the greatest global threat facing the world over the next two years is “misinformation and disinformation.” One challenge for making sense of the WEF’s warning and how risks today may differ from the sorts of misinformation historically endemic to politics is that the notion of “misinformation” lacks a meaningful definition, by the WEF or by misinformation scholars.

For instance, in 2015, one group of scholars explained:

Misinformation by definition does not accurately reflect the true state of the world.

Simple, right? Once we identify “truth” then we will know that everything else is misinformation. But how do we know what is actually true? And who is “we,” anyway?

A 2023 survey of misinformation researchers offered little further help in defining misinformation, explaining,

the most popular definition of misinformation was “false and misleading information”

That definition is about as useful as defining research misconduct as improper conduct in research — circular and empty.

Some have suggested that we can use expert consensus views to delineate truth from misinformation, but this too is deeply problematic and leads to challenging questions such as:

Consider a highly visible debate of the past few days over climate science. One position, expressed by climate scientist James Hansen, holds that the views of the Intergovernmental Panel on Climate Change (IPCC) on the pace of ongoing climate change are deeply flawed and subject to the myopia of its authors — Hansen believes that climate change is proceeding much faster than projected by the IPCC.

Does Hansen’s stance — explicitly contrary to the IPCC consensus reports — make him a peddler of misinformation?

An opposing view, expressed by climate scientist Michael E. Mann, is that Hansen’s views are “absolutely absurd” and the IPCC is correct in its views on the pace of evolving climate change.3

Hansen is the doyen of climate science and activism, so does that make Mann’s divergent views misinformation?

Who holds the truth here? Who is engaged in misinformation?

The answer to both questions is — Neither!

Important scientific questions with relevance to policy and politics are often contested, with legitimate views expressed that are in opposition or collectively inconsistent. Post-normal science is in fact perfectly normal.

Contestation of truth claims among experts does not mean that we know nothing or that anyone’s version of reality is just as good as anyone else’s version. It does mean that truth often sits in abeyance, meaning that there are typically and simultaneously multiple valid truth claims — contested certainties as Steve Rayner used to call them. Such excess of objectivity unavoidably provides fertile ground for cherry-picking and building a case for vastly different perspectives and policy perspectives.4

Our collective inability to achieve omniscience is one important reason why the management of experts and expert knowledge in the context of democratic governance is so very important for the sustainability of both science and democracy.5

Of course, the Earth is round, the vast majority of vaccines are safe and effective, and climate change is real. In political debates, opponents on competing sides of an issues often invoke these sorts of tame issues to suggest that the much larger class of wicked problems are just as simple, and can be handed the same way. This is how the notions of science and consensus are often weaponized to stifle debate and discussion under the guise of protecting ordinary people from “misinformation” — but dealing with climate change is not at all like responding to an approaching tornado.

Misinformation experts analogize misinformation to a disease — a removable stone in the head. They do not prescribe drilling a hole in the skull to remove the idiocy, however, they do assert that they are akin to medical doctors who can “inoculate” the misguided and susceptible against misinformation, giving them occasional “boosters” and with the goal of creating “herd immunity.”

The self-serving nature of such arguments is hard to miss, and it is easy to see why governments, social media platforms, major news media, and advocacy groups might find such arguments appealing in service of their own interests.6

Source; “Vaccine for Brainwash” in Psychology Today, McGuire 1970, cited in Lewandowsky and van der Linden (2021)

For instance, Lewandowsky and van der Linden (2021) explain their theory as follows:

Just as vaccines are weakened versions of a pathogen that trigger the production of antibodies when they enter the body to help confer immunity against future infection, inoculation theory postulates that the same can be achieved with information: by preemptively exposing people to a sufficiently weakened version of a persuasive attack, a cognitive-motivational process is triggered that is analogous to the production of “mental antibodies”, rendering the individual more immune to persuasion.

It is of course well established that authorities can manipulate public perceptions and beliefs. Almost a century ago, political scientist Harold Lasswell in his PhD dissertation defined propaganda as “the management of collective attitudes by the manipulation of significant symbols.” Propaganda has a pejorative connotation, but under Lasswell’s definition everyone engaged in political discourse is engaged in propaganda.7 You can’t swim without getting wet.

Can misinformation experts influence the opinions and expressed views of study subjects in experimental settings? Of course they can.

Can experts manipulate opinions outside such settings? It is more difficult, but of course we can. “Inoculation” is a word that is more inscrutable and palatable than propaganda, but it means the same thing — the management of collective attitudes.

Contrary to the views of many misinformation scholars, there is of course absolutely no problem with experts calling things like they see them.8 For instance, climate scientists James Hansen and Michael Mann have every right to express their divergent opinions — whether they are popular or unpopular, in the fat or thin part of a scientific consensus, or perceived as supporting this or that political agenda. Similarly, there is considerable value in teaching people critical thinking skills, fallacies of logic and reasoning, and how to avoid being fooled or exploited.

Problems arise when experts collaborate with authoritative institutions to enforce a party line perspective on views that are deemed acceptable (and unacceptable) to express and that party line is then used to (try to) manage public opinion and expert discourse. Party lines can be enforced by institutions of the media, of government, and of course of science. An obvious recent example has been the ongoing effort to manage discourse over the possibility that COVID-19 resulted from a research-related incident. That effort has failed, but we are still dealing with the consequences.

Party line perspectives need not be the result of explicit decisions, but rather can result from like-mindedness and groupthink. Consider the self-reported political leaning of more than 150 misinformation researchers, as shown below. Can it be surprising that the “inoculations” (propaganda) applied by these scholars to wrong-thinking citizens have a decidedly leftward tilt?

Source: X/Twitter, based on data from the Harvard Misinformation Review, graph created by Bjorn Lomborg with comments from Sander van der Linden.

The past several decades has seen the pathological politicization of science across several important fields, climate and public health among them. Much of this is of course due to the fact that these issues are deeply politically contested with partisan combatants on the left and right (who are otherwise at odds on policy) in strong agreement that they can achieve political gains through polarization of these issues.

While politicians and advocates bear their share of responsibility, make no mistake, we in the expert community have also played our part in pathological politicization by becoming partisans as well — especially authoritative scientific institutions and their leaders. Misinformation research and its role in seeking to enforce party line perspectives are but one example of how experts have chosen to fuel the flames of polarization.

Fortunately for both science and democracy, there are much better alternatives than seeking to remove the “stone of madness.”



This archive was generated by a fusion of Pipermail (Mailman edition) and MHonArc.